Coresets For Monotonic Functions with Applications to Deep Learning

نویسندگان

  • Elad Tolochinsky
  • Dan Feldman
چکیده

Coreset (or core-set) in this paper is a small weighted subset Q of the input set P with respect to a given monotonic function f : R → R that provably approximates its fitting loss ∑ p∈P f(p · x) to any given x ∈ R. Using Q we can obtain approximation to x∗ that minimizes this loss, by running existing optimization algorithms on Q. We provide: (i) a lower bound that proves that there are sets with no coresets smaller than n = |P | , (ii) a proof that a small coreset of size near-logarithmic in n exists for any input P , under natural assumption that holds e.g. for logistic regression and the sigmoid activation function. (iii) a generic algorithm that computes Q in O(nd+n logn) expected time, (iv) novel technique for improving existing deep networks using such coresets, (v) extensive experimental results with open code.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Maximum Margin Coresets for Active and Noise Tolerant Learning

We study the problem of learning large margin halfspaces in various settings using coresets and show that coresets are a widely applicable tool for large margin learning. A large margin coreset is a subset of the input data sufficient for approximating the true maximum margin solution. In this work, we provide a direct algorithm and analysis for constructing large margin coresets. We show vario...

متن کامل

Deep Lattice Networks and Partial Monotonic Functions

We propose learning deep models that are monotonic with respect to a userspecified set of inputs by alternating layers of linear embeddings, ensembles of lattices, and calibrators (piecewise linear functions), with appropriate constraints for monotonicity, and jointly training the resulting network. We implement the layers and projections with new computational graph nodes in TensorFlow and use...

متن کامل

Taming the Waves: Sine as Activation Function in Deep Neural Networks

Most deep neural networks use non-periodic and monotonic—or at least quasiconvex— activation functions. While sinusoidal activation functions have been successfully used for specific applications, they remain largely ignored and regarded as difficult to train. In this paper we formally characterize why these networks can indeed often be difficult to train even in very simple scenarios, and desc...

متن کامل

A novel method based on a combination of deep learning algorithm and fuzzy intelligent functions in order to classification of power quality disturbances in power systems

Automatic classification of power quality disturbances is the foundation to deal with power quality problem. From the traditional point of view, the identification process of power quality disturbances should be divided into three independent stages: signal analysis, feature selection and classification. However, there are some inherent defects in signal analysis and the procedure of manual fe...

متن کامل

Improving Stock Return Forecasting by Deep Learning Algorithm

Improving return forecasting is very important for both investors and researchers in financial markets. In this study we try to aim this object by two new methods. First, instead of using traditional variable, gold prices have been used as predictor and compare the results with Goyal's variables. Second, unlike previous researches new machine learning algorithm called Deep learning (DP) has bee...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1802.07382  شماره 

صفحات  -

تاریخ انتشار 2018